[8.x](backport #5614) Adjust memory requests and limits for elastic-agent when run in Kubernetes cluster #5657
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does this PR do?
This PR adjusts the default memory requests and limits for elastic-agent when it runs in a Kubernetes cluster.
The new requests and limits will be
These will be adequate even in big load (tested with 95 pods per node).
Why is it important?
In latest elastic-agent versions the baseline memory consumption is higher, thus adding Kubernetes and system
integration leads to often OOM killed pods.
The previous memory limit set was 700Mb
After investigation with various number of pods we have seen that in ^8.15.* versions
the memory consumption can reach from 750 to 950 Mb.
This varies depending on the number of pods a single elastic agent has to monitor.
For example:
Checklist
./changelog/fragments
using the changelog toolDisruptive User Impact
The inadequate memory limits lead to elastic-agent getting often OOM killed in newer versions.
How to test this PR locally
Deploy Elastic Agent with system and Kubernetes integration following the Kibana instructions and watch the
memory consumption of the pod with
watch "kubectl top pod -n kube-system"
and its status.Related issues
This is an automatic backport of pull request #5614 done by [Mergify](https://mergify.com).